Goto

Collaborating Authors

 design automation conference



LithoBench: Benchmarking AI Computational Lithography for Semiconductor Manufacturing Supplementary Materials

Neural Information Processing Systems

It also incorporates Python programs that can train and test the models mentioned in this paper. By inheriting the classes, users can easily build their own models that can be trained and tested by LithoBench, without the need of writing the code for data loading and evaluation. For average pooling, we use a kernel size of 7 and a stride of 1. PyTorch builtin functions so that an SGD optimizer with a learning rate of 0.5 can be used to optimize Table 1 compares the performance of our reference IL T algorithm with SOT A IL T algorithms. We provide the PNG images of the all data. The connections between adjacent vertices are horizontal or vertical. In this section, we describe the details of the DNN models used in this paper.




LithoBench: Benchmarking AI Computational Lithography for Semiconductor Manufacturing Supplementary Materials

Neural Information Processing Systems

It also incorporates Python programs that can train and test the models mentioned in this paper. By inheriting the classes, users can easily build their own models that can be trained and tested by LithoBench, without the need of writing the code for data loading and evaluation. For average pooling, we use a kernel size of 7 and a stride of 1. PyTorch builtin functions so that an SGD optimizer with a learning rate of 0.5 can be used to optimize Table 1 compares the performance of our reference IL T algorithm with SOT A IL T algorithms. We provide the PNG images of the all data. The connections between adjacent vertices are horizontal or vertical. In this section, we describe the details of the DNN models used in this paper.



TRACE: Learning to Compute on Graphs

Zheng, Ziyang, Zhu, Jiaying, Zhou, Jingyi, Xu, Qiang

arXiv.org Artificial Intelligence

Learning to compute, the ability to model the functional behavior of a computational graph, is a fundamental challenge for graph representation learning. Yet, the dominant paradigm is architecturally mismatched for this task. This flawed assumption, central to mainstream message passing neural networks (MPNNs) and their conventional Transformer-based counterparts, prevents models from capturing the position-aware, hierarchical nature of computation. To resolve this, we introduce \textbf{TRACE}, a new paradigm built on an architecturally sound backbone and a principled learning objective. First, TRACE employs a Hierarchical Transformer that mirrors the step-by-step flow of computation, providing a faithful architectural backbone that replaces the flawed permutation-invariant aggregation. Second, we introduce \textbf{function shift learning}, a novel objective that decouples the learning problem. Instead of predicting the complex global function directly, our model is trained to predict only the \textit{function shift}, the discrepancy between the true global function and a simple local approximation that assumes input independence. We validate this paradigm on electronic circuits, one of the most complex and economically critical classes of computational graphs. Across a comprehensive suite of benchmarks, TRACE substantially outperforms all prior architectures. These results demonstrate that our architecturally-aligned backbone and decoupled learning objective form a more robust paradigm for the fundamental challenge of learning to compute on graphs.


Alignment Unlocks Complementarity: A Framework for Multiview Circuit Representation Learning

Shi, Zhengyuan, Wang, Jingxin, Jiang, Wentao, Ma, Chengyu, Zheng, Ziyang, Chu, Zhufei, Qian, Weikang, Xu, Qiang

arXiv.org Artificial Intelligence

Multiview learning on Boolean circuits holds immense promise, as different graph-based representations offer complementary structural and semantic information. However, the vast structural heterogeneity between views--such as an And-Inverter Graph (AIG) versus an XOR-Majority Graph (XMG)--poses a critical barrier to effective fusion, especially for self-supervised techniques like masked modeling. Naively applying such methods fails, as the cross-view context is perceived as noise. Our key insight is that functional alignment is a necessary precondition to unlock the power of multiview self-supervision. We introduce MixGate, a framework built on a principled training curriculum that first teaches the model a shared, function-aware representation space via an Equivalence Alignment Loss. Only then do we introduce a multiview masked modeling objective, which can now leverage the aligned views as a rich, complementary signal. Extensive experiments, including a crucial ablation study, demonstrate that our alignment-first strategy transforms masked modeling from an ineffective technique into a powerful performance driver. Multiview learning on Boolean circuits holds immense promise, as different graph-based representations offer complementary structural and semantic insights. While an And-Inverter Graph (AIG) provides a detailed structural view, a format like an XOR-Majority Graph (XMG) offers a semantically richer, high-level abstraction. This multiview approach has shown remarkable empirical success, surpassing earlier models that relied on single representations Li et al. (2022); Wang et al. (2022); Wu et al. (2023); Shi et al. (2023); Deng et al. (2024); Wang et al. (2024). The key challenge, however, arises from the vast structural heterogeneity between these views.


Automated Multi-Agent Workflows for RTL Design

Bhattaram, Amulya, Ramamoorthy, Janani, Gupta, Ranit, Marculescu, Diana, Stamoulis, Dimitrios

arXiv.org Artificial Intelligence

The rise of agentic AI workflows unlocks novel opportunities for computer systems design and optimization. However, for specialized domains such as program synthesis, the relative scarcity of HDL and proprietary EDA resources online compared to more common programming tasks introduces challenges, often necessitating task-specific fine-tuning, high inference costs, and manually-crafted agent orchestration. In this work, we present VeriMaAS, a multi-agent framework designed to automatically compose agentic workflows for RTL code generation. Our key insight is to integrate formal verification feedback from HDL tools directly into workflow generation, reducing the cost of gradient-based updates or prolonged reasoning traces. Our method improves synthesis performance by 5-7% for pass@k over fine-tuned baselines, while requiring only a few hundred training examples, representing an order-of-magnitude reduction in supervision cost.


AnalogCoder-Pro: Unifying Analog Circuit Generation and Optimization via Multi-modal LLMs

Lai, Yao, Poddar, Souradip, Lee, Sungyoung, Chen, Guojin, Hu, Mengkang, Yu, Bei, Luo, Ping, Pan, David Z.

arXiv.org Artificial Intelligence

Despite recent advances, analog front-end design still relies heavily on expert intuition and iterative simulations, which limits the potential for automation. We present AnalogCoder-Pro, a multimodal large language model (LLM) framework that integrates generative and optimization techniques. The framework features a multimodal diagnosis-and-repair feedback loop that uses simulation error messages and waveform images to autonomously correct design errors. It also builds a reusable circuit tool library by archiving successful designs as modular subcircuits, accelerating the development of complex systems. Furthermore, it enables end-to-end automation by generating circuit topologies from target specifications, extracting key parameters, and applying Bayesian optimization for device sizing. On a curated benchmark suite covering 13 circuit types, AnalogCoder-Pro successfully designed 28 circuits and consistently outperformed existing LLM-based methods in figures of merit.